Sparse Variations and Reducibility among Prediction Problems Authors :

نویسنده

  • Naoki Abe
چکیده

We investigate the relationship between two di erent notions of reducibility among prediction (learning) problems within the distribution-free learning model of Valiant (PAC learning model [7, 2].) The notions of reducibility we consider are the analogues for prediction problems of the many-one reducibility and of the Turing reducibility. The former is the notion of prediction preserving reducibility developed by Pitt and Warmuth [6], and its generalization. Concerning these two notions of reducibility, we show that there exist a pair of prediction problems A and B, whose membership problems are polynomial time solvable, such that A is reducible to B with respect to the Turing reducibility, but not with respect to the prediction preserving reducibility. We show this result by making use of the notion of a class of polynomially sparse variants of a concept representation class. We rst show that any class A of polynomially sparse variants of another class B is reducible to B with respect to the `Turing reducibility.' We then prove the existence of a prediction problem R and a class R 0 of polynomially sparse variants of R, such that R 0 does not reduce to R with respect to the prediction preserving reducibility.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Face Recognition in Thermal Images based on Sparse Classifier

Despite recent advances in face recognition systems, they suffer from serious problems because of the extensive types of changes in human face (changes like light, glasses, head tilt, different emotional modes). Each one of these factors can significantly reduce the face recognition accuracy. Several methods have been proposed by researchers to overcome these problems. Nonetheless, in recent ye...

متن کامل

Relating Equivalence and Reducibility to Sparse Sets

For various polynomial-time reducibilities r, this paper asks whether being r-reducible to a sparse set is a broader notion than being r-equivalent to a sparse set. Although distinguishing equivalence and reducibility to sparse sets, for many-one or 1-truth-table reductions, would imply that P 6= NP, this paper shows that for k-truth-table reductions, k 2, equivalence and reducibility to sparse...

متن کامل

Mammalian Eye Gene Expression Using Support Vector Regression to Evaluate a Strategy for Detecting Human Eye Disease

Background and purpose: Machine learning is a class of modern and strong tools that can solve many important problems that nowadays humans may be faced with. Support vector regression (SVR) is a way to build a regression model which is an incredible member of the machine learning family. SVR has been proven to be an effective tool in real-value function estimation. As a supervised-learning appr...

متن کامل

Reductions among prediction problems: on the difficulty of predicting automata

Consider the following question: Given examples of words accepted and rejected by an unknown automaton, is there an algorithm that in a feasible amount of time will learn to predict which words will be accepted by the automaton? We develop a notion of prediction-preserving reducibility, and show that if DFAs are predictable, then so are all languages in logspace. In particular, the predictabili...

متن کامل

On Reductions to Sets that Avoid EXPSPACE

A set B is called EXPSPACE-avoiding, if every subset of B in EXPSPACE is sparse. For example, sets of high information density (called HIGH sets in 5]) are shown to be EXPSPACE-avoiding. Investigating the complexity of sets A in EXPSPACE that honestly reduce to EXPSPACE-avoiding sets, we show that if the reducibility used has a property, called range-constructibility, then A must also reduce to...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1992